robust controller
Forthcoming machine learning and AI seminars: December 2025 edition
This post contains a list of the AI-related seminars that are scheduled to take place between 1 December 2025 and 31 January 2026. All events detailed here are free and open for anyone to attend virtually. Dick den Hertog (University of Amsterdam) Association of European Operational Research Societies To receive the seminar link, sign up to the mailing list . Annabelle Gawer The Digital Humanism (DIGHUM) Initiative The talk will be livestreamed on YouTube here . Jesús Moreno-León (University of Seville) Raspberry PI Sign up here to join.
- Europe > Netherlands > North Holland > Amsterdam (0.25)
- North America > United States > Virginia (0.05)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.05)
- (2 more...)
AIhub monthly digest: November 2025 – learning robust controllers, trust in multi-agent systems, and a new fairness evaluation dataset
Welcome to our monthly digest, where you can catch up with any AIhub stories you may have missed, peruse the latest news, recap recent events, and more. This month, we learn about rewarding explainability in drug repurposing with knowledge graphs, investigate value-aligned autonomous vehicles, and consider trust in multi-agent systems. In this blog post, and write about work, presented at the International Joint Conference on Artificial Intelligence (IJCAI2025), on rewarding explainability in drug repurposing with knowledge graphs. Their work introduces a reinforcement learning approach that not only predicts which drug-disease pairs might hold promise but also explains why. Astrid Rakow writes about designing "conflict-sensitive" autonomous traffic agents that explicitly recognise, reason about, and act upon competing ethical, legal, and social values.
- North America > United States (0.30)
- Europe > Netherlands > North Holland > Amsterdam (0.05)
- Europe > Italy > Emilia-Romagna > Metropolitan City of Bologna > Bologna (0.05)
A Robust Controller based on Gaussian Processes for Robotic Manipulators with Unknown Uncertainty
Giacomuzzo, Giulio, Abdelwahab, Mohamed, Calì, Marco, Libera, Alberto Dalla, Carli, Ruggero
In this paper, we propose a novel learning-based robust feedback linearization strategy to ensure precise trajectory tracking for an important family of Lagrangian systems. We assume a nominal knowledge of the dynamics is given but no a-priori bounds on the model mismatch are available. In our approach, the key ingredient is the adoption of a regression framework based on Gaussian Processes (GPR) to estimate the model mismatch. This estimate is added to the outer loop of a classical feedback linearization scheme based on the nominal knowledge available. Then, to compensate for the residual uncertainty, we robustify the controller including an additional term whose size is designed based on the variance provided by the GPR framework. We proved that, with high probability, the proposed scheme is able to guarantee asymptotic tracking of a desired trajectory. We tested numerically our strategy on a 2 degrees of freedom planar robot.
A Parameter Adaptive Trajectory Tracking and Motion Control Framework for Autonomous Vehicle
Song, Jiarui, Sun, Yingbo, Dong, Qing, Ji, Xuewu
This paper studies the trajectory tracking and motion control problems for autonomous vehicles (AVs). A parameter adaptive control framework for AVs is proposed to enhance tracking accuracy and yaw stability. While establishing linear quadratic regulator (LQR) and three robust controllers, the control framework addresses trajectory tracking and motion control in a modular fashion, without introducing complexity into each controller. The robust performance has been guaranteed in three robust controllers by considering the parameter uncertainties, mismatch of unmodeled subsystem as well as external disturbance, comprehensively. Also, the dynamic characteristics of uncertain parameters are identified by Recursive Least Squares (RLS) algorithm, while the boundaries of three robust factors are determined through combining Gaussian Process Regression (GPR) and Bayesian optimization machine learning methods, reducing the conservatism of the controller. Sufficient conditions for closed-loop stability under the diverse robust factors are provided by the Lyapunov method analytically. The simulation results on MATLAB/Simulink and Carsim joint platform demonstrate that the proposed methodology considerably improves tracking accuracy, driving stability, and robust performance, guaranteeing the feasibility and capability of driving in extreme scenarios.
- Automobiles & Trucks (1.00)
- Energy (0.89)
- Transportation > Ground > Road (0.68)
Robust Cooperative Multi-Agent Reinforcement Learning:A Mean-Field Type Game Perspective
Zaman, Muhammad Aneeq uz, Laurière, Mathieu, Koppel, Alec, Başar, Tamer
In this paper, we study the problem of robust cooperative multi-agent reinforcement learning (RL) where a large number of cooperative agents with distributed information aim to learn policies in the presence of \emph{stochastic} and \emph{non-stochastic} uncertainties whose distributions are respectively known and unknown. Focusing on policy optimization that accounts for both types of uncertainties, we formulate the problem in a worst-case (minimax) framework, which is is intractable in general. Thus, we focus on the Linear Quadratic setting to derive benchmark solutions. First, since no standard theory exists for this problem due to the distributed information structure, we utilize the Mean-Field Type Game (MFTG) paradigm to establish guarantees on the solution quality in the sense of achieved Nash equilibrium of the MFTG. This in turn allows us to compare the performance against the corresponding original robust multi-agent control problem. Then, we propose a Receding-horizon Gradient Descent Ascent RL algorithm to find the MFTG Nash equilibrium and we prove a non-asymptotic rate of convergence. Finally, we provide numerical experiments to demonstrate the efficacy of our approach relative to a baseline algorithm.
- Asia > China > Shanghai > Shanghai (0.04)
- North America > United States > Illinois > Champaign County > Urbana (0.04)
Data-Guided Regulator for Adaptive Nonlinear Control
Rahimi, Niyousha, Mesbahi, Mehran
A critical aspect of autonomous operations in safety-critical scenarios is learning from available data for quick adaptation to new environments while maintaining safety. Examples include aircraft emergency landing scenarios in adverse weather conditions and agile quadrotor flights through low clearance gates in the presence of dynamic and strong wind conditions [1]. From a system theoretic perspective, this system feature maps to having the autonomous agent handle parametric model uncertainties and disturbances with control-theoretic guarantees such as stability and tracking error convergence, common in adaptive control settings [2, 3]. A rich body of literature has analyzed classical adaptive control algorithms' stability and convergence properties for continuous-time dynamical systems. Such studies include the use of PI (proportional integral) controllers [4] for a class of linear time-varying systems to guarantee (I) infinite-time convergence of the tracking error to zero, i.e., the difference between actual and nominal states () = () (), for any constant exogenous disturbance (denoted by), (II) infinite-time convergence of the tracking error () to a bound which is proportional to the bound on the magnitude of the rate of the exogenous signal ().
- North America > United States > Washington > King County > Seattle (0.14)
- North America > United States > Pennsylvania > Philadelphia County > Philadelphia (0.04)
Experience Transfer for Robust Direct Data-Driven Control
von Rohr, Alexander, Likhachev, Dmitrii, Trimpe, Sebastian
Learning-based control uses data to design efficient controllers for specific systems. When multiple systems are involved, experience transfer usually focuses on data availability and controller performance yet neglects robustness to variations between systems. In contrast, this letter explores experience transfer from a robustness perspective. We leverage the transfer to design controllers that are robust not only to the uncertainty regarding an individual agent's model but also to the choice of agent in a fleet. Experience transfer enables the design of safe and robust controllers that work out of the box for all systems in a heterogeneous fleet. Our approach combines scenario optimization and recent formulations for direct data-driven control without the need to estimate a model of the system or determine uncertainty bounds for its parameters. We demonstrate the benefits of our data-driven robustification method through a numerical case study and obtain learned controllers that generalize well from a small number of open-loop trajectories in a quadcopter simulation.
- Information Technology > Knowledge Management > Knowledge Engineering (1.00)
- Information Technology > Artificial Intelligence > Robots (1.00)
- Information Technology > Artificial Intelligence > Machine Learning (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (0.34)
Improving the Performance of Robust Control through Event-Triggered Learning
von Rohr, Alexander, Solowjow, Friedrich, Trimpe, Sebastian
Robust controllers ensure stability in feedback loops designed under uncertainty but at the cost of performance. Model uncertainty in time-invariant systems can be reduced by recently proposed learning-based methods, which improve the performance of robust controllers using data. However, in practice, many systems also exhibit uncertainty in the form of changes over time, e.g., due to weight shifts or wear and tear, leading to decreased performance or instability of the learning-based controller. We propose an event-triggered learning algorithm that decides when to learn in the face of uncertainty in the LQR problem with rare or slow changes. Our key idea is to switch between robust and learned controllers. For learning, we first approximate the optimal length of the learning phase via Monte-Carlo estimations using a probabilistic model. We then design a statistical test for uncertain systems based on the moment-generating function of the LQR cost. The test detects changes in the system under control and triggers re-learning when control performance deteriorates due to system changes. We demonstrate improved performance over a robust controller baseline in a numerical example.
- Europe > Germany > North Rhine-Westphalia > Cologne Region > Aachen (0.04)
- Europe > Germany > Baden-Württemberg > Stuttgart Region > Stuttgart (0.04)
- Asia > Middle East > Jordan (0.04)
Sim-to-Real: Learning Agile Locomotion For Quadruped Robots
Tan, Jie, Zhang, Tingnan, Coumans, Erwin, Iscen, Atil, Bai, Yunfei, Hafner, Danijar, Bohez, Steven, Vanhoucke, Vincent
Designing agile locomotion for quadruped robots often requires extensive expertise and tedious manual tuning. In this paper, we present a system to automate this process by leveraging deep reinforcement learning techniques. Our system can learn quadruped locomotion from scratch using simple reward signals. In addition, users can provide an open loop reference to guide the learning process when more control over the learned gait is needed. The control policies are learned in a physics simulator and then deployed on real robots. In robotics, policies trained in simulation often do not transfer to the real world. We narrow this reality gap by improving the physics simulator and learning robust policies. We improve the simulation using system identification, developing an accurate actuator model and simulating latency. We learn robust controllers by randomizing the physical environments, adding perturbations and designing a compact observation space. We evaluate our system on two agile locomotion gaits: trotting and galloping. After learning in simulation, a quadruped robot can successfully perform both gaits in the real world.
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.04)
- Information Technology > Artificial Intelligence > Robots > Locomotion (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.93)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Undirected Networks > Markov Models (0.46)
Circumventing Robots' Failures by Embracing Their Faults: A Practical Approach to Planning for Autonomous Construction
Witwicki, Stefan (Swiss Federal Institute of Technology (EPFL)) | Mondada, Francesco (Swiss Federal Institute of Technology (EPFL))
This paper overviews our application of state-of-the-art automated planning algorithms to real mobile robots performing an autonomous construction task, a domain in which robots are prone to faults. We describe how embracing these faults leads to better representations and smarter planning, allowing robots with limited precision to avoid catastrophic failures and succeed in intricate constructions.